Interpretability via Model Extraction

نویسندگان

  • Osbert Bastani
  • Carolyn Kim
  • Hamsa Bastani
چکیده

Œe ability to interpret machine learning models has become increasingly important now that machine learning is used to inform consequential decisions. We propose an approach called model extraction for interpreting complex, blackbox models. Our approach approximates the complex model using a much more interpretable model; as long as the approximation quality is good, then statistical properties of the complex model are reƒected in the interpretable model. We show how model extraction can be used to understand and debug random forests and neural nets trained on several datasets from the UCI Machine Learning Repository, as well as control policies learned for several classical reinforcement learning problems.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Mind the Gap: A Generative Approach to Interpretable Feature Selection and Extraction

We present the Mind the Gap Model (MGM), an approach for interpretable feature extraction and selection. By placing interpretability criteria directly into the model, we allow for the model to both optimize parameters related to interpretability and to directly report a global set of distinguishable dimensions to assist with further data exploration and hypothesis generation. MGM extracts disti...

متن کامل

Learning Tensor-Based Features for Whole-Brain fMRI Classification

This paper presents a novel tensor-based feature learning approach for whole-brain fMRI classification. Whole-brain fMRI data have high exploratory power, but they are challenging to deal with due to large numbers of voxels. A critical step for fMRI classification is dimensionality reduction, via feature selection or feature extraction. Most current approaches perform voxel selection based on f...

متن کامل

SnapToGrid: From Statistical to Interpretable Models for Biomedical Information Extraction

We propose an approach for biomedical information extraction that marries the advantages of machine learning models, e.g., learning directly from data, with the benefits of rule-based approaches, e.g., interpretability. Our approach starts by training a feature-based statistical model, then converts this model to a rule-based variant by converting its features to rules, and “snapping to grid” t...

متن کامل

Interpretability and learning in neuro-fuzzy systems

A methodology for the development of linguistically interpretable fuzzy models from data is presented. The implementation of the model is conducted through the training of a neuro-fuzzy network, i.e., a neural net architecture capable of representing a fuzzy system. In the /rst phase, the structure of the model is obtained by means of subtractive clustering, which allows the extraction of a set...

متن کامل

A Rule Extractor for Diagnosing the Type 2 Diabetes Using a Self-organizing Genetic Algorithm

Introduction: Constructing medical decision support models to automatically extract knowledge from data helps physicians in early diagnosis of disease. Interpretability of the inferential rules of these models is a key indicator in determining their performance in order to understand how they make decisions, and increase the reliability of their output. Methods: In this study, an automated hyb...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • CoRR

دوره abs/1706.09773  شماره 

صفحات  -

تاریخ انتشار 2017